翻訳と辞書
Words near each other
・ C3orf15
・ C3orf23
・ C3orf58
・ C3orf60
・ C3orf70
・ C4
・ C4 (band)
・ C4 (classification)
・ C4 (Colombia)
・ C4 (conference)
・ C4 (rapper)
・ C4 (TV channel)
・ C4 antisense RNA
・ C4 carbon fixation
・ C4 Engine
C4.5 algorithm
・ C40
・ C40 Cities Climate Leadership Group
・ C40H44N4O16
・ C40H54O2
・ C40H56
・ C40H56O
・ C40H56O2
・ C40H56O3
・ C40H56O4
・ C40H58
・ C40H64
・ C41
・ C41H26O26
・ C41H28O26


Dictionary Lists
翻訳と辞書 辞書検索 [ 開発暫定版 ]
スポンサード リンク

C4.5 algorithm : ウィキペディア英語版
C4.5 algorithm

C4.5 is an algorithm used to generate a decision tree developed by Ross Quinlan.〔Quinlan, J. R. C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers, 1993.〕 C4.5 is an extension of Quinlan's earlier ID3 algorithm. The decision trees generated by C4.5 can be used for classification, and for this reason, C4.5 is often referred to as a statistical classifier.
It became quite popular after ranking #1 in the ''Top 10 Algorithms in Data Mining'' pre-eminent paper published by Springer LNCS in 2008.〔(Umd.edu - Top 10 Algorithms in Data Mining )〕
==Algorithm==
C4.5 builds decision trees from a set of training data in the same way as ID3, using the concept of information entropy. The training data is a set S = of already classified samples. Each sample s_i consists of a p-dimensional vector (x_, x_, ...,x_) , where the x_j represent attribute values or features of the sample, as well as the class in which s_i falls.
At each node of the tree, C4.5 chooses the attribute of the data that most effectively splits its set of samples into subsets enriched in one class or the other. The splitting criterion is the normalized information gain (difference in entropy). The attribute with the highest normalized information gain is chosen to make the decision. The C4.5 algorithm then recurs on the smaller sublists.
This algorithm has a few base cases.
*All the samples in the list belong to the same class. When this happens, it simply creates a leaf node for the decision tree saying to choose that class.
*None of the features provide any information gain. In this case, C4.5 creates a decision node higher up the tree using the expected value of the class.
*Instance of previously-unseen class encountered. Again, C4.5 creates a decision node higher up the tree using the expected value.

抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)
ウィキペディアで「C4.5 algorithm」の詳細全文を読む



スポンサード リンク
翻訳と辞書 : 翻訳のためのインターネットリソース

Copyright(C) kotoba.ne.jp 1997-2016. All Rights Reserved.